MASS: Multi-Attentional Semantic Segmentation of LiDAR Data for Dense Top-View Understanding

نویسندگان

چکیده

At the heart of all automated driving systems is ability to sense surroundings, e.g., through semantic segmentation LiDAR sequences, which experienced a remarkable progress due release large datasets such as SemanticKITTI and nuScenes-LidarSeg. While most previous works focus on xmlns:xlink="http://www.w3.org/1999/xlink">sparse input, xmlns:xlink="http://www.w3.org/1999/xlink">dense output masks provide self-driving cars with almost complete environment information. In this paper, we introduce MASS - Multi-Attentional Semantic Segmentation model specifically built for dense top-view understanding scenes. Our framework operates pillar- occupancy features comprises three attention-based building blocks: (1) keypoint-driven graph attention, (2) an LSTM-based attention computed from vector embedding spatial (3) pillar-based resulting in 360° mask. With extensive experiments both, nuScenes-LidarSeg, quantitatively demonstrate effectiveness our model, outperforming state art by 19.0% reaching 30.4% mIoU where first work addressing task. Furthermore, multi-attention shown be very effective 3D object detection validated KITTI-3D dataset, showcasing its high generalizability other tasks related vision.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Understanding Convolution for Semantic Segmentation

Recent advances in deep learning, especially deep convolutional neural networks (CNNs), have led to significant improvement over previous semantic segmentation systems. Here we show how to improve pixel-wise semantic segmentation by manipulating convolution-related operations that are of both theoretical and practical value. First, we design dense upsampling convolution (DUC) to generate pixel-...

متن کامل

Learning Dense Convolutional Embeddings for Semantic Segmentation

This paper proposes a new deep convolutional neural network (DCNN) architecture for learning semantic segmentation. The main idea is to train the DCNN to produce internal representations that respect object boundaries. That is, for any two pixels on the same object, the DCNN is trained to produce nearly-identical internal representations; conversely, the DCNN is trained to produce dissimilar re...

متن کامل

Learning Where to Classify in Multi-view Semantic Segmentation

There is an increasing interest in semantically annotated 3D models, e.g. of cities. The typical approaches start with the semantic labelling of all the images used for the 3D model. Such labelling tends to be very time consuming though. The inherent redundancy among the overlapping images calls for more efficient solutions. This paper proposes an alternative approach that exploits the geometry...

متن کامل

Learning Multi-level Region Consistency with Dense Multi-label Networks for Semantic Segmentation

Semantic image segmentation is a fundamental task in image understanding. Per-pixel semantic labelling of an image benefits greatly from the ability to consider region consistency both locally and globally. However, many Fully Convolutional Network based methods do not impose such consistency, which may give rise to noisy and implausible predictions. We address this issue by proposing a dense m...

متن کامل

Dense Multi-view 3D-reconstruction Without Dense Correspondences

We introduce a variational method for multi-view shapefrom-shading under natural illumination. The key idea is to couple PDE-based solutions for single-image based shapefrom-shading problems across multiple images and multiple color channels by means of a variational formulation. Rather than alternatingly solving the individual SFS problems and optimizing the consistency across images and chann...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE Transactions on Intelligent Transportation Systems

سال: 2022

ISSN: ['1558-0016', '1524-9050']

DOI: https://doi.org/10.1109/tits.2022.3145588